A Distance for HMMs Based on Aggregated Wasserstein Metric and State Registration
نویسندگان
چکیده
We propose a framework, named Aggregated Wasserstein, for computing a dissimilarity measure or distance between two Hidden Markov Models with state conditional distributions being Gaussian. For such HMMs, the marginal distribution at any time position follows a Gaussian mixture distribution, a fact exploited to softly match, aka register, the states in two HMMs. We refer to such HMMs as Gaussian mixture model-HMM (GMM-HMM). The registration of states is inspired by the intrinsic relationship of optimal transport and the Wasserstein metric between distributions. Specifically, the components of the marginal GMMs are matched by solving an optimal transport problem where the cost between components is the Wasserstein metric for Gaussian distributions. The solution of the optimization problem is a fast approximation to the Wasserstein metric between two GMMs. The new Aggregated Wasserstein distance is a semi-metric and can be computed without generating Monte Carlo samples. It is invariant to relabeling or permutation of states. The distance is defined meaningfully even for two HMMs that are estimated from data of different dimensionality, a situation that can arise due to missing variables. This distance quantifies the dissimilarity of GMM-HMMs by measuring both the difference between the two marginal GMMs and that between the two transition matrices. Our new distance is tested on tasks of retrieval, classification, and t-SNE visualization of time series. Experiments on both synthetic and real data have demonstrated its advantages in terms of accuracy as well as efficiency in comparison with existing distances based on the Kullback-Leibler divergence.
منابع مشابه
Convergence in the Wasserstein Metric for MarkovChain
This paper gives precise bounds on the convergence time of the Gibbs sampler used in the Bayesian restoration of a degraded image. Convergence to stationarity is assessed using the Wasserstein metric, rather than the usual choice of total variation distance. The Wasserstein metric may be more easily applied in some applications, particularly those on continuous state spaces. Bounds on convergen...
متن کاملLearning with a Wasserstein Loss
Learning to predict multi-label outputs is challenging, but in many problems there is a natural metric on the outputs that can be used to improve predictions. In this paper we develop a loss function for multi-label learning, based on the Wasserstein distance. The Wasserstein distance provides a natural notion of dissimilarity for probability measures. Although optimizing with respect to the ex...
متن کاملMulti-scale Non-Rigid Point Cloud Registration Using Robust Sliced-Wasserstein Distance via Laplace-Beltrami Eigenmap
In this work, we propose computational models and algorithms for point cloud registration with non-rigid transformation. First, point clouds sampled from manifolds originally embedded in some Euclidean space RD are transformed to new point clouds embedded in Rn by Laplace-Beltrami(LB) eigenmap using the n leading eigenvalues and corresponding eigenfunctions of LB operator defined intrinsically ...
متن کاملA note on reinforcement learning with Wasserstein distance regularisation, with applications to multipolicy learning
In this note we describe an application of Wasserstein distance to Reinforcement Learning. The Wasserstein distance in question is between the distribution of mappings of trajectories of a policy into some metric space, and some other fixed distribution (which may, for example, come from another policy). Different policies induce different distributions, so given an underlying metric, the Wasse...
متن کاملCommon fixed points for a pair of mappings in $b$-Metric spaces via digraphs and altering distance functions
In this paper, we discuss the existence and uniqueness of points of coincidence and common fixed points for a pair of self-mappings satisfying some generalized contractive type conditions in $b$-metric spaces endowed with graphs and altering distance functions. Finally, some examples are provided to justify the validity of our results.
متن کامل